Goto

Collaborating Authors

 Coast Province


Detecting Network-based Internet Censorship via Latent Feature Representation Learning

Duncan, Shawn P., Chen, Hui

arXiv.org Artificial Intelligence

Internet censorship is a phenomenon of societal importance and attracts investigation from multiple disciplines. Several research groups, such as Censored Planet, have deployed large scale Internet measurement platforms to collect network reachability data. However, existing studies generally rely on manually designed rules (i.e., using censorship fingerprints) to detect network-based Internet censorship from the data. While this rule-based approach yields a high true positive detection rate, it suffers from several challenges: it requires human expertise, is laborious, and cannot detect any censorship not captured by the rules. Seeking to overcome these challenges, we design and evaluate a classification model based on latent feature representation learning and an image-based classification model to detect network-based Internet censorship. To infer latent feature representations fromnetwork reachability data, we propose a sequence-to-sequence autoencoder to capture the structure and the order of data elements in the data. To estimate the probability of censorship events from the inferred latent features, we rely on a densely connected multi-layer neural network model. Our image-based classification model encodes a network reachability data record as a gray-scale image and classifies the image as censored or not using a dense convolutional neural network. We compare and evaluate both approaches using data sets from Censored Planet via a hold-out evaluation. Both classification models are capable of detecting network-based Internet censorship as we were able to identify instances of censorship not detected by the known fingerprints. Latent feature representations likely encode more nuances in the data since the latent feature learning approach discovers a greater quantity, and a more diverse set, of new censorship instances.


Geographic Citation Gaps in NLP Research

Rungta, Mukund, Singh, Janvijay, Mohammad, Saif M., Yang, Diyi

arXiv.org Artificial Intelligence

In a fair world, people have equitable opportunities to education, to conduct scientific research, to publish, and to get credit for their work, regardless of where they live. However, it is common knowledge among researchers that a vast number of papers accepted at top NLP venues come from a handful of western countries and (lately) China; whereas, very few papers from Africa and South America get published. Similar disparities are also believed to exist for paper citation counts. In the spirit of "what we do not measure, we cannot improve", this work asks a series of questions on the relationship between geographical location and publication success (acceptance in top NLP venues and citation impact). We first created a dataset of 70,000 papers from the ACL Anthology, extracted their meta-information, and generated their citation network. We then show that not only are there substantial geographical disparities in paper acceptance and citation but also that these disparities persist even when controlling for a number of variables such as venue of publication and sub-field of NLP. Further, despite some steps taken by the NLP community to improve geographical diversity, we show that the disparity in publication metrics across locations is still on an increasing trend since the early 2000s. We release our code and dataset here: https://github.com/iamjanvijay/acl-cite-net


Facebook uses satellite imagery machine learning and AI

#artificialintelligence

Facebook uses satellite imagery machine learning and AI to prepare maps for locating unconnected communities across the world. Maps tell us so much more than how to get from A to B, or where C is in relation to D. They can be tools of power and snapshots of history; they can give urban planners the information to plan infrastructure. After a disaster, population density and crisis maps help to direct aid and aid workers. Throughout time, different cultures and industries have produced radically different images of the world. Today there are more than 7 billion humans sprawling across Earth.


When Disaster Strikes, He Creates A 'Crisis Map' That Helps Save Lives

NPR Technology

Patrick Meier (center, in cap) flies a drone in Nepal after the earthquake in 2015. Meier and his team were able to to capture detailed images of damage around the capital, Kathmandu. He believes using this technology will make crisis mapping even more effective for disaster response. Patrick Meier (center, in cap) flies a drone in Nepal after the earthquake in 2015. Meier and his team were able to to capture detailed images of damage around the capital, Kathmandu. He believes using this technology will make crisis mapping even more effective for disaster response.


Facebook's AI team maps Earth to beam internet access to all

#artificialintelligence

Social networking giant Facebook is using its artificial intelligence (AI) technology and resources to map the entire Earth and launch the world's most detailed population maps that will help it beam cheap internet to remote areas. To begin with, the Facebook AI team crunched 14.6 billion images of maps from across 20 countries, including India, covering 21.6 million sq kms to come up with the first detailed map of human settlement for these countries. "This is an impressive project from our team developing solar-powered planes for beaming down internet connectivity and our AI research team. Many people live in remote communities and accurate data on where people live doesn't always exist," wrote Facebook CEO Mark Zuckerberg in a latest post. The 20 countries mapped were Algeria, Burkina Faso, Cameroon, Egypt, Ethiopia, Ghana, India, Ivory Coast, Kenya, Madagascar, Mexico, Mozambique, Nigeria, South Africa, Sri Lanka, Tanzania, Turkey, Uganda, Ukraine and Uzbekistan.